Parallelism in Structured Newton Computations
نویسندگان
چکیده
A fundamental computational procedure in practically all areas of scientific computing is the calculation of the Newton step (in n-dimensions). In many cases this computation represents the dominant cost in the overall computing task. Typically the Newton step computation breaks down into two separable subtasks: calculation of the Jacobian (or Hessian) matrix along with the right-hand-side, and then the solution of a linear system (which, in turn, may involve a matrix factorization). Both subtasks can be expensive though in many problems it is the first, calculation of the function and derivative matrices, that dominates. In most cases when the Newton step computation is relatively expensive, the function that yields the Newton system is itself a ‘structured’ computation. A structured computation is one that breaks down into a (partially ordered) straight-line sequence of (accessible) macro computational subtasks. For example, if F is a function that is computed by evaluating the sequence F1, F2, F3, in order, then F is a structured computation. The general structured situation can be described as follows: F is a structured computation, z = F (x), if F is evaluated by computing a (partially-ordered) sequence of intermediate vectors y defined below: Solve for y1 : F 1 (x, y1) = 0 Solve for y2 : F 2 (x, y1, y2) = 0 .. .. Solve for yp : F p (x, y1, y2, · · · , yp) = 0 “Solve” for output z : z − F p+1(x, y1, y2, · · · , yp) = 0 . (1)
منابع مشابه
John von Neumann Institute for Computing Parallelism in Structured Newton Computations
c © 2007 by John von Neumann Institute for Computing Permission to make digital or hard copies of portions of this work for personal or classroom use is granted provided that the copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise requires prior specific permission by the publisher ment...
متن کاملPerformance implications of synchronization structure in parallel programming
The restricted synchronization structure of so-called structured parallel programming paradigms has an advantageous effect on programmer productivity, cost modeling, and scheduling complexity. However, imposing these restrictions can lead to a loss of parallelism, compared to using a programming approach that does not impose synchronization structure. In this paper we study the potential loss o...
متن کاملOn the convergence speed of artificial neural networks in the solving of linear systems
Artificial neural networks have the advantages such as learning, adaptation, fault-tolerance, parallelism and generalization. This paper is a scrutiny on the application of diverse learning methods in speed of convergence in neural networks. For this aim, first we introduce a perceptron method based on artificial neural networks which has been applied for solving a non-singula...
متن کاملNew Approach for the Inversion of Structured Matrices via Newton’s Iteration
Newton’s iteration is a fundamental tool for numerical solutions of systems of equations. The well-known iteration ( ) 1 2 , 0 i i i X X I MX i + = − ≥ rapidly refines a crude initial approximation 0 X to the inverse of a general nonsingular matrix. In this paper, we will extend and apply this method to n n × structured matrices M , in which matrix multiplication has a lower computational cost....
متن کاملScalable solution of non-linear time-dependent systems
We study parallel solution methods for time-dependent problems where the domain decomposition is in the time variable. We show that Newton’s method can be used to create parallelism in nonlinear dynamical systems. We demonstrate that arbitrarily large speedups can be expected for sufficiently small time steps. Thus we conclude that nonlinear dynamical systems do not form essentially sequential ...
متن کامل